AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Continuous Pretraining Optimization

# Continuous Pretraining Optimization

Finance Llama3 8B
A financial domain model developed based on Llama3-8B, enhanced with an instruction pretraining framework to improve domain adaptability, achieving or even surpassing the performance of Llama3-70B on financial tasks.
Large Language Model Transformers English
F
instruction-pretrain
1,200
64
Layerskip Llama2 7B
Other
An improved model based on Llama2 7B, supporting hierarchical skip and self-speculative decoding to enhance inference efficiency
Large Language Model Transformers English
L
facebook
1,674
14
Quietstar 8 Ahead
Based on the Mistral-7b model, it employs the Quiet-STaR method for continuous pretraining, generating 8 reasoning tokens before each output token to enhance reasoning capabilities.
Large Language Model Transformers
Q
ezelikman
239
90
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase